Multimodal Human-Agent Communication

نویسنده

  • David Rosén
چکیده

In this paper I will present my approach on how to merge the HCI field with the new growing field of agent technology, especially from the point of view of the use and need of a personal assistant. Multimodality is a must if we are ever to implement a user-friendly, intuitive and effective interface for a user to interact with a system (agent or other). Selecting modality for presentation can be done by mapping the values of their characteristics onto each other and setting up constraints to select the one most appropriate during runtime. When designing a system the same constraints can be set up to find what kind of multimodality that should be supported to allow a user to enter the input that the user wants or that the system requires. When allowing a user to interact with a system using several modalities there is also a need to synchronize and interpret the input, which can be done by merging input from different modalities based on their temporal relation and the expected input. As with all technologies, especially new ones, standards is an important issue. FIPA (Foundation for Intelligent Physical Agents) is working on several areas within this field, both in creating an agent communication language (ACL) and a standard for human-agent communication. The PIM project (Personal Information Management system) here at Telia Research AB, Software Tech. is an attempt to determine the needs in a personal assistant. Even though its just a prototype, with some minor extensions to the system, it will act as the base for this paper when determining the need of modalities and architecture. The need of modalities for this system is heavily based on speech. In most cases a simple word spotter is enough but when dictating a letter over the phone, a more sophisticated natural language recognizer is needed. Another modality that is needed is the pen based gestures and written text, this is invaluable when interacting with the graphical interface to the users calendar. To choose an architecture for this particular type of application was not a very difficult task. The OAA has been used in similar applications and it follows the basic notion on being a system of cooperating agents. Nevertheless it still lacks some parts in the multimodal communication support, e.g. user modeling, which is one thing that can not go unmentioned. Therefore I have outlined the basics of my …

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Achieving Multimodal Cohesion during Intercultural Conversations

How do English as a lingua franca (ELF) speakers achieve multimodal cohesion on the basis of their specific interests and cultural backgrounds? From a dialogic and collaborative view of communication, this study focuses on how verbal and nonverbal modes cohere together during intercultural conversations. The data include approximately 160-minute transcribed video recordings of ELF interactions ...

متن کامل

Langage de conversation multimodal pour agent conversationnel animé

This article falls within the realm of dialogue between a human and an Embodied Conversational Agent (ECA). We claim that a specific agent conversational language is needed for such interactions, based on the essential role of emotion in human communication. In order to define this language, we propose a library of Multimodal Conversation Acts is proposed, based in particular on speech acts and...

متن کامل

Multimodal Temporal Patterns for the Analysis of User’s Involvement in Affective Interaction with Virtual Agents

Given the growing interest in developing embodied virtual agents with multimodal communication and emotional expression abilities, the issue of user’s involvement is a relevant topic to take into account in determining how to assess and interpret the quality of user-agent affective interaction. Main goal of this paper is the definition of a methodology for the analysis of user-agent interaction...

متن کامل

Communicating and Acting: Understanding Gesture in Simulation Semantics

In this paper, we introduce an architecture for multimodal communication between humans and computers engaged in a shared task. We describe a representative dialogue between an artificial agent and a human that will be demonstrated live during the presentation. This assumes a multimodal environment and semantics for facilitating communication and interaction with a computational agent. To this ...

متن کامل

Actes de langage et émotions : vers un langage de conversation multimodal

Agent Communication languages were initially developed for exchanging information within multiagent systems composed of software entities. This article falls within the realm of mixed Multi-Agent Systems composed of humans and artificial agents which are embodied for simulating the human behavior. In order to increase the interactional capacity of virtual agents, a multimodal conversation acts ...

متن کامل

Kommunikative Rhythmen in Gestik und Sprache

Led by the fundamental role that rhythms apparently play in speech and gestural communication among humans, this study was undertaken to substantiate a biologically motivated model for synchronizing speech and gesture input in human computer interaction. Our approach presents a novel method which conceptualizes a multimodal user interface on the basis of timed agent systems. We use multiple age...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1998